Lecture 9 : Decision - making under total uncertainty : the multiplicative weight algorithm
نویسنده
چکیده
(Today’s notes below are largely lifted with minor modifications from a survey by Arora, Hazan, Kale in Theory of Computing journal, Volume 8 (2012), pp. 121-164.) Today we study decision-making under total uncertainty: there is no a priori distribution on the set of possible outcomes. (This line will cause heads to explode among devout Bayesians, but it makes sense in many computer science settings. One reason is computational complexity or general lack of resources: the decision-maker usually lacks the computational power to construct the tree of all exp(T ) outcomes possible in the next T steps, and the resources to do enough samples/polls/surveys to figure out their distribution. Or the algorithm designer may not be a Bayesian.) Such decision-making (usually done with efficient algorithms) is studied in the field of online computation, which takes the view that the algorithm is responding to a sequence of requests that arrive one by one. The algorithm must take an action as each request arrives, and it may discover later, after seeing more requests, that its past actions were suboptimal. But past actions cannot be unchanged. See the book by Borodin and El-Yaniv for a fuller introduction to online algorithms. This lecture and the next covers one such success story: an online optimization tool called the multiplicative weight update method. The power of the method arises from the very minimalistic assumptions, which allow it to be plugged into various settings (as we will do in next lecture).
منابع مشابه
Lecture 8 : Decision - making under total uncertainty : the multiplicative weight algorithm
(Today’s notes below are largely lifted with minor modifications from a survey by Arora, Hazan, Kale in Theory of Computing journal, Volume 8 (2012), pp. 121-164.) Today we study decision-making under total uncertainty: there is no a priori distribution on the set of possible outcomes. (This line will cause heads to explode among devout Bayesians, but it makes sense in many computer science set...
متن کاملA New Compromise Decision-making Model based on TOPSIS and VIKOR for Solving Multi-objective Large-scale Programming Problems with a Block Angular Structure under Uncertainty
This paper proposes a compromise model, based on a new method, to solve the multi-objective large-scale linear programming (MOLSLP) problems with block angular structure involving fuzzy parameters. The problem involves fuzzy parameters in the objective functions and constraints. In this compromise programming method, two concepts are considered simultaneously. First of them is that the optimal ...
متن کاملUNCERTAINTY DATA CREATING INTERVAL-VALUED FUZZY RELATION IN DECISION MAKING MODEL WITH GENERAL PREFERENCE STRUCTURE
The paper introduces a new approach to preference structure, where from a weak preference relation derive the following relations:strict preference, indifference and incomparability, which by aggregations and negations are created and examined. We decomposing a preference relation into a strict preference, anindifference, and an incomparability relation.This approach allows one to quantify diff...
متن کاملOptimal Cropping Pattern Modifications with the Aim of Environmental-Economic Decision Making Under Uncertainty
Sustainability in agricultural is determined by aspects like economy, society and environment. Multi-objective programming (MOP) model has been a widely used tool for studying and analyzing the sustainability of agricultural system. However, optimization models in most applications are forced to use data which is uncertain. Recently, robust optimization has been used as an optimization model th...
متن کاملCS261: A Second Course in Algorithms Lecture #12: Applications of Multiplicative Weights to Games and Linear Programs∗
1 Extensions of the Multiplicative Weights Guarantee Last lecture we introduced the multiplicative weights algorithm for online decision-making. You don't need to remember the algorithm details for this lecture, but you should remember that it's a simple and natural algorithm (just one simple update per action per time step). You should also remember its regret guarantee, which we proved last l...
متن کامل